Authors: Taifeng Liu, Yang Liu, Zhuo Ma, Tong Yang, Xinjing Liu, Teng Li, Jianfeng Ma
Class: EE 7700 ML for CPS
Instructor: Dr. Xugui Zhou
Presented By: Group 4 on 11/04/2025
Summarized By: Group 3
L-HAWK: A Controllable Physical Adversarial Patch Against a Long-Distance
Target (NDSS 2025) introduces a new kind of laser-triggered physical attack on
autonomous vehicles (AVs).
Conventional adversarial patches fool AV vision models but are always active
and affect every vehicle nearby. L-HAWK overcomes this by using a printed
patch that stays harmless until a laser signal activates it,
allowing attackers to target a specific vehicle at up to 50 meters.
The authors propose an asynchronous learning framework that jointly optimizes both the patch and laser parameters under real-world noise, distances, and lighting. Experiments on object detectors (YOLO, Faster-R-CNN) and classifiers (VGG, ResNet, Inception) show attack success rates over 90%, outperforming previous methods by more than 50% and extending attack distance sevenfold.
Key contributions include:
● A controllable long-range physical adversarial patch activated by laser signals.
● A new trigger modeling and joint optimization method combining physical and digital learning.
● Empirical validation in both digital and physical setups with high robustness.
Limitations involve dependency on camera frame timing, safety constraints during testing, and reduced stealth under night or close-range conditions.

This is the introduction slide of the presentation.
This slide shows the highlights of the presentation and what are the things that are going to be discussed.

This Slide contains the name and Authors of the presentation and the venue where it was presented.

In this slide, we see the description of the title and the general idea of this study.

Here, we can see the motivation of this work and why it is important. ML models have been used for perception models and in safety critical scenarios. It can be used in autonomous vehicles and in different other sectors. These types of models are vulnerable to physical perturbations and can be fooled which can be dangerous. There are different forms of patch attack, one can generate patches digitally and by means of a physical method.

Here we can see the attack and defense models which highlight the types of attacks and methods of defense as well as examples and where L-Hawk falls in place.

This slide shows the visual recognition model in Autonomous Driving and how it can be fooled using Adversarial Patches. These patches can be generated physically or can be generated using adversarial attacks on the Neural network.

In this slide, we can see the conventional adversarial patch attack that is generated on stop signs. It affects all the passing vehicles that are trying to pick up stop sign signals from camera.
Slide - 9: Research Question

This is the main research question that this paper is trying to solve.

This slide presents several existing studies addressing this research question and the specific problems they aimed to solve. These works explore the use of acoustic signals, laser-colored signals, electromagnetic interference, and triggered physical adversarial patch methods to attack autonomous driving models.

This work proposes a laser-based attack that obstructs the vehicle’s camera, preventing it from capturing the correct road-sign image. It also shows that the laser method outperforms ultrasonic attacks because it remains effective at longer ranges.

This is a key slide explaining how the laser can be controlled to obscure the specific pixels that contain the object in the camera’s field of view. During camera exposure the laser emits short, localized pulses; because rolling-shutter sensors record rows sequentially, those pulses can affect particular image rows. Controllable laser parameters include wavelength, power, pulse width, aim angle, and distance, while the resulting appearance also depends on the camera settings.

Here we can see how the laser can selectively affect vehicles when the object (road sign) appears in their camera exposure.

This slide shows the paper’s primary goal: to optimize a printable adversarial patch and a controllable laser trigger so the attack can be precisely controlled.

This slide explains how real-world laser noise differs from synthetic noise used in simulations. Real laser-induced patterns are structured and asymmetric due to rolling-shutter and lens scattering effects. Modeling this empirical noise produces camera responses much closer to real images, significantly reducing MSE compared to generic noise models.

They build a library of real laser triggers by recording frames across power/distance/angle/light, isolating transient stripes via frame-differences and masking, and indexing each trigger for robust sampling during optimization.

This slide explains the core challenge: jointly optimizing a printable adversarial patch (continuous, pixel-space) and a controllable laser trigger (discrete, physical) creates a coupled digital and physical problem that is unstable under direct joint training. The solution is an alternating optimization strategy that updates the digital patch and the physical trigger separately while constraining triggers to the measured trigger space.

Here we can see how to update the digital patch with the current trigger, then refine trigger parameters, and repeat until both converge to a stable, controllable attack.

Here we can see four types of attacks: hiding attack, creating attack, targeted attack against detector and targeted attack against classifier. We will know about all of this in the later description.

The evaluation involves the main experiment setup which evaluates four attack types and shows high attack success rates (≈90–99%) across multiple detectors and classifiers for different attack types.

They showed video demonstrations in this slide of the attack testing how distance, position, and lighting affect the attack. They demonstrate that L-HAWK remains effective beyond 25m with strong robustness.

Outdoor experiments up to 30m and at varying incidence angles with three laser powers were also conducted. Quantitative results confirmed improved stability at longer ranges and higher power levels, with attack success above 80% under moderate lighting. Higher power beyond 20m however yields diminishing returns.

This slide summarizes the four main contributions in their own words namely: controllable laser-triggered patch, novel asynchronous optimization, investigation of laser parameters, and digital and physical validation.

In their own words, they note three drawbacks: reliance on knowing camera refresh rate, limited evaluation scale, and reduced stealth at close or dark conditions.

This slide links the paper to CPS security by showing the attack targets the ML perception in autonomous vehicles, a core cyber-physical system component.

This slide introduces how later studies build on L-HAWK and the evolving research directions in physical adversarial perception.

They show the current landscape by highlighting works on physical patches, optical, and electromagnetic triggers. They also show L-HAWK’s place at the intersection of controllable and long-distance methods.

This slide lists Poltergeist, Rolling Colors, GlitchHiker, and TPatch as earlier works that inspired controllable or optical adversarial attacks and how they went about it.

This slide explains TPatch’s acoustic trigger mechanism and how L-HAWK extends it using a laser for longer range and better control.

They describe a prior laser-based color-stripe attack that misleads sign perception at a distance. L-HAWK builds upon its rolling-shutter insights to learn realistic triggers.

The emerging directions after L-Hawk lean towards future work in near-infrared (NIR) attacks and multimodal (radar/LiDAR) triggers for more stealth and robustness.

Recent research using NIR light for hidden attacks at night extends L-HAWK’s ideas to invisible spectrums. The key idea is to model an attack in a spectrum common to ADAS sensors. It also suggests the use of multi-spectrum triggers.

They detail each member’s role in this slide with Jordan handling literature review and formatting, and Shafikul focusing on methodology and evaluation analysis.

This slide introduced a segment for questions and answers which stimulated discussions about feasibility and implications and the impact L-Hawk has for the future of AVs.

The discussion slide opened another interactive section of the presentation inviting class debate on the questions in the following slides.

The class also discussed whether L-HAWK-type attacks could bypass multi-sensor fusion e.g., LiDAR + cameras and how adaptive defense mechanisms might randomize camera exposure to block such attacks.
Future triggers could move beyond lasers and sound into multi-spectrum or cross-modal triggers, such as infrared or electromagnetic interference invisible to humans but detectable by sensors. Researchers may also explore sensor-fusion deception, where coordinated signals subtly mislead multiple sensing modalities simultaneously for more reliable, stealthy activation.
Our group believes Glare and autofocus adjustments most strongly degrade L-HAWK’s effectiveness by changing camera exposure and suppressing the laser-induced stripe. To counter this, attackers could use adaptive power control or synchronized pulsing that compensates for brightness changes and maintains consistent pixel intensity on the sensor.

Lists all cited academic works, including Poltergeist, Rolling Colors, TPatch, and NIR-based attacks.

Closing slide with acknowledgments and end of presentation.